Download AWS Certified Big Data-Specialty Dumps.AWS-Certified-Big-Data-Specialty.ExamTopics.2025-05-22.65q.vcex

Vendor: Amazon
Exam Code: AWS-Certified-Big-Data-Specialty
Exam Name: AWS Certified Big Data-Specialty Dumps
Date: May 22, 2025
File Size: 58 KB
Downloads: 2

How to open VCEX files?

Files with VCEX extension can be opened by ProfExam Simulator.

ProfExam Discount

Demo Questions

Question 1
An Amazon Redshift Database is encrypted using KMS. A data engineer needs to use the AWS CLI to create a KMS encrypted snapshot of the database in another AWS region.
Which three steps should the data engineer take to accomplish this task? (Choose three.)
  1. Create a new KMS key in the destination region.
  2. Copy the existing KMS key to the destination region.
  3. Use CreateSnapshotCopyGrant to allow Amazon Redshift to use the KMS key from the source region.
  4. In the source region, enable cross-region replication and specify the name of the copy grant created.
  5. In the destination region, enable cross-region replication and specify the name of the copy grant created.
  6. Use CreateSnapshotCopyGrant to allow Amazon Redshift to use the KMS key created in the destination region. ADF
Correct answer: Explanation
Question 2
A company has several teams of analysts. Each team of analysts has their own cluster. The teams need to run SQL queries using Hive, Spark-SQL, and Presto with Amazon EMR. The company needs to enable a centralized metadata layer to expose the Amazon S3 objects as tables to the analysts.
Which approach meets the requirement for a centralized metadata layer?
  1. EMRFS consistent view with a common Amazon DynamoDB table
  2. Bootstrap action to change the Hive Metastore to an Amazon RDS database
  3. s3distcp with the outputManifest option to generate RDS DDL
  4. Naming scheme support with automatic partition discovery from Amazon S3
Correct answer: A
Question 3
An organization is developing a mobile social application and needs to collect logs from all devices on which it is installed. The organization is evaluating the Amazon Kinesis Data Streams to push logs and Amazon EMR to process data. They want to store data on HDFS using the default replication factor to replicate data among the cluster, but they are concerned about the durability of the data. Currently, they are producing 300 GB of raw data daily, with additional spikes during special events. They will need to scale out the Amazon EMR cluster to match the increase in streamed data.
Which solution prevents data loss and matches compute demand?
  1. Use multiple Amazon EBS volumes on Amazon EMR to store processed data and scale out the Amazon EMR cluster as needed.
  2. Use the EMR File System and Amazon S3 to store processed data and scale out the Amazon EMR cluster as needed.
  3. Use Amazon DynamoDB to store processed data and scale out the Amazon EMR cluster as needed.
  4. use Amazon Kinesis Data Firehose and, instead of using Amazon EMR, stream logs directly into Amazon Elasticsearch Service.
Correct answer: D
Question 4
A company receives data sets coming from external providers on Amazon S3. Data sets from different providers are dependent on one another. Data sets will arrive at different times and in no particular order.
A data architect needs to design a solution that enables the company to do the following:
  • Rapidly perform cross data set analysis as soon as the data becomes available
  • Manage dependencies between data sets that arrive at different times
Which architecture strategy offers a scalable and cost-effective solution that meets these requirements?
  1. Maintain data dependency information in Amazon RDS for MySQL. Use an AWS Data Pipeline job to load an Amazon EMR Hive table based on task dependencies and event notification triggers in Amazon S3.
  2. Maintain data dependency information in an Amazon DynamoDB table. Use Amazon SNS and event notifications to publish data to fleet of Amazon EC2 workers. Once the task dependencies have been resolved, process the data with Amazon EMR.
  3. Maintain data dependency information in an Amazon ElastiCache Redis cluster. Use Amazon S3 event notifications to trigger an AWS Lambda function that maps the S3 object to Redis. Once the task dependencies have been resolved, process the data with Amazon EMR.
  4. Maintain data dependency information in an Amazon DynamoDB table. Use Amazon S3 event notifications to trigger an AWS Lambda function that maps the S3 object to the task associated with it in DynamoDB. Once all task dependencies have been resolved, process the data with Amazon EMR.
Correct answer: C
Question 5
An organization needs to store sensitive information on Amazon S3 and process it through Amazon EMR. Data must be encrypted on Amazon S3 and Amazon EMR at rest and in transit. Using Thrift Server, the Data Analysis team users HIVE to interact with this data. The organization would like to grant access to only specific databases and tables, giving permission only to the SELECT statement.
Which solution will protect the data and limit user access to the SELECT statement on a specific portion of data?
  1. Configure Transparent Data Encryption on Amazon EMR. Create an Amazon EC2 instance and install Apache Ranger. Configure the authorization on the cluster to use Apache Ranger.
  2. Configure data encryption at rest for EMR File System (EMRFS) on Amazon S3. Configure data encryption in transit for traffic between Amazon S3 and EMRFS. Configure storage and SQL base authorization on HiveServer2.
  3. Use AWS KMS for encryption of data. Configure and attach multiple roles with different permissions based on the different user needs.
  4. Configure Security Group on Amazon EMR. Create an Amazon VPC endpoint for Amazon S3. Configure HiveServer2 to use Kerberos authentication on the cluster.
Correct answer: C
Question 6
A Redshift data warehouse has different user teams that need to query the same table with very different query types. These user teams are experiencing poor performance.
Which action improves performance for the user teams in this situation?
  1. Create custom table views.
  2. Add interleaved sort keys per team.
  3. Maintain team-specific copies of the table.
  4. Add support for workload management queue hopping.
Correct answer: D
Question 7
An organization's data warehouse contains sales data for reporting purposes. data governance policies prohibit staff from accessing the customers' credit card numbers.
How can these policies be adhered to and still allow a Data Scientist to group transactions that use the same credit card number?
  1. Store a cryptographic hash of the credit card number.
  2. Encrypt the credit card number with a symmetric encryption key, and give the key only to the authorized Data Scientist.
  3. Mask the credit card numbers to only show the last four digits of the credit card number.
  4. Encrypt the credit card number with an asymmetric encryption key and give the decryption key only to the authorized Data Scientist.
Correct answer: C
Question 8
An administrator needs to design the event log storage architecture for events from mobile devices. The event data will be processed by an Amazon EMR cluster daily for aggregated reporting and analytics before being archived.
How should the administrator recommend storing the log data?
  1. Create an Amazon S3 bucket and write log data into folders by device. Execute the EMR job on the device folders.
  2. Create an Amazon DynamoDB table partitioned on the device and sorted on date, write log data to table. Execute the EMR job on the Amazon DynamoDB table.
  3. Create an Amazon S3 bucket and write data into folders by day. Execute the EMR job on the daily folder.
  4. Create an Amazon DynamoDB table partitioned on EventID, write log data to table. Execute the EMR job on the table.
Correct answer: A
Question 9
A data engineer wants to use an Amazon Elastic Map Reduce for an application. The data engineer needs to make sure it complies with regulatory requirements. The auditor must be able to confirm at any point which servers are running and which network access controls are deployed.
Which action should the data engineer take to meet this requirement?
  1. Provide the auditor IAM accounts with the SecurityAudit policy attached to their group.
  2. Provide the auditor with SSH keys for access to the Amazon EMR cluster.
  3. Provide the auditor with CloudFormation templates.
  4. Provide the auditor with access to AWS DirectConnect to use their existing tools.
Correct answer: C
Question 10
A gaming organization is developing a new game and would like to offer real-time competition to their users. The data architecture has the following characteristics:
  • The game application is writing events directly to Amazon DynamoDB from the user's mobile device.
  • Users from the website can access their statistics directly from DynamoDB.
  • The game servers are accessing DynamoDB to update the user's information.
  • The data science team extracts data from DynamoDB for various applications.
The engineering team has already agreed to the IAM roles and policies to use for the data science team and the application.
Which actions will provide the MOST security, while maintaining the necessary access to the website and game application? (Choose two.)
  1. Use Amazon Cognito user pool to authenticate to both the website and the game application.
  2. Use IAM identity federation to authenticate to both the website and the game application.
  3. Create an IAM policy with PUT permission for both the website and the game application.
  4. Create an IAM policy with fine-grained permission for both the website and the game application.
  5. Create an IAM policy with PUT permission for the game application and an IAM policy with GET permission for the website.
Correct answer: BE
Question 11
An Amazon EMR cluster using EMRFS has access to petabytes of data on Amazon S3, originating from multiple unique data sources. The customer needs to query common fields across some of the data sets to be able to perform interactive joins and then display results quickly.
Which technology is most appropriate to enable this capability?
  1. Presto
  2. MicroStrategy
  3. Pig
  4. R Studio
Correct answer: C
HOW TO OPEN VCE FILES

Use VCE Exam Simulator to open VCE files
Avanaset

HOW TO OPEN VCEX AND EXAM FILES

Use ProfExam Simulator to open VCEX and EXAM files
ProfExam Screen

ProfExam
ProfExam at a 20% markdown

You have the opportunity to purchase ProfExam at a 20% reduced price

Get Now!